Attention Flows: Analyzing and Comparing Attention Mechanisms in Language Models
نویسندگان
چکیده
Advances in language modeling have led to the development of deep attention-based models that are performant across a wide variety natural processing (NLP) problems. These typified by pre-training process on large unlabeled text corpora and subsequently fine-tuned for specific tasks. Although considerable work has been devoted understanding attention mechanisms pre-trained models, it is less understood how model's change when trained target NLP task. In this paper, we propose visual analytics approach fine-tuning models. Our visualization, Attention Flows, designed support users querying, tracing, comparing within layers, amongst heads Transformer-based To help gain insight classification decision made, our design centered depicting classification-based at deepest layer from prior layers flows throughout words input. Flows supports analysis single model, as well comparison between via their similarities differences. We use study various sentence tasks highlight evolves address nuances solving these
منابع مشابه
Comparing the Relationship between Focus of Attention and Attention Bias in Suicide and Non-suicide Attempters
Objective: Suicide rate is unfortunately increasing in recent years and its assessment is a must. Most studies have evaluated different factors like personality, depression, or environment in suicide attempters but cognitive functions have been overlooked in suicide attempters. The current research aimed to evaluate a kind of cognitive function, namely attention in survivors of suicide. ...
متن کاملAttention and Visual Perception: Analyzing Divided Attention
Attention is one of the most active areas of experimental research in cognitive science. This chapter focuses on an aspect of attention that has been studied intensively over the past twenty-five years or so: capacity limitations on our ability to perceive multiple visual inputs. The basic questions are simple but far-reaching. How much visual information can we take in at one time? What can we...
متن کاملCoherent Dialogue with Attention-Based Language Models
We model coherent conversation continuation via RNNbased dialogue models equipped with a dynamic attention mechanism. Our attention-RNN language model dynamically increases the scope of attention on the history as the conversation continues, as opposed to standard attention (or alignment) models with a fixed input scope in a sequence-tosequence model. This allows each generated word to be assoc...
متن کاملMechanisms and Representations of Language-Mediated Visual Attention
The experimental investigation of language-mediated visual attention is a promising way to study the interaction of the cognitive systems involved in language, vision, attention, and memory. Here we highlight four challenges for a mechanistic account of this oculomotor behavior: the levels of representation at which language-derived and vision-derived representations are integrated; attentional...
متن کاملcomparing the relationship between focus of attention and attention bias in suicide and non-suicide attempters
objective: suicide rate is unfortunately increasing in recent years and its assessment is a must. most studies have evaluated different factors like personality, depression, or environment in suicide attempters but cognitive functions have been overlooked in suicide attempters. the current research aimed to evaluate a kind of cognitive function, namely attention in survivors of suicide. method...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
ژورنال
عنوان ژورنال: IEEE Transactions on Visualization and Computer Graphics
سال: 2021
ISSN: ['1077-2626', '2160-9306', '1941-0506']
DOI: https://doi.org/10.1109/tvcg.2020.3028976